Multiagent Coordination in Cooperative Q-learning Systems
نویسندگان
چکیده
Many reinforcement learning architectures fail to learn optimal group behaviors in the multiagent domain. Although these coordination difficulties are often attributed to the non-Markovian environment created by the gradually-changing policies of concurrently learning agents, a careful analysis of the situation reveals an underlying problem structure which can cause suboptimal group policies even when the Markovian properties of the learning environment are preserved. This underlying structure is termed the multiagent coordination problem, and it can be viewed as a combination of two related but distinct limitations of cooperative multiagent learning systems: action shadowing and joint action prediction. This paper discusses the causes of each of these limitations and their effects on systems of cooperative Q-learning agents, including the conditions which must be met in order to guarantee the execution of optimal group policies. Multiagent coordination strategies presented by other researchers are considered from the perspective of this new problem framework, and an algorithm is presented which extends some of these coordination strategies to improve the tractability of large-scale multiagent learning.
منابع مشابه
Predicting and Preventing Coordination Problems in Cooperative Q-learning Systems
We present a conceptual framework for creating Qlearning-based algorithms that converge to optimal equilibria in cooperative multiagent settings. This framework includes a set of conditions that are sufficient to guarantee optimal system performance. We demonstrate the efficacy of the framework by using it to analyze several well-known multi-agent learning algorithms and conclude by employing i...
متن کاملBaselines for Joint-Action Reinforcement Learning of Coordination in Cooperative Multi-agent Systems
We report on an investigation of reinforcement learning techniques for the learning of coordination in cooperative multiagent systems. Specifically, we focus on a novel action selection strategy for Q-learning (Watkins 1989). The new technique is applicable to scenarios where mutual observation of actions is not possible. To date, reinforcement learning approaches for such independent agents di...
متن کاملImproving on the reinforcement learning of coordination in cooperative multi-agent systems
We report on an investigation of reinforcement learning techniques for the learning of coordination in cooperative multiagent systems. These techniques are variants of Q-learning (Watkins, 1989) that are applicable to scenarios where mutual observation of actions is not possible. To date, reinforcement learning approaches for such independent agents did not guarantee convergence to the optimal ...
متن کاملMultiagent Q-Learning by Context-Specific Coordination Graphs
One of the main problems in cooperative multiagent learning is that the joint action space is exponential in the number of agents. In this paper, we investigate a sparse representation of the joint action space in which value rules specify the coordination dependencies between the different agents for a particular state. Each value rule has an associated payoff which is part of the global Q-fun...
متن کاملReinforcement social learning of coordination in cooperative multiagent systems
Coordination in cooperative multiagent systems is an important problem and has received a lot of attention in multiagent learning literature. Most of previous works study the problem of how two (or more) players can coordinate on Pareto-optimal Nash equilibrium(s) through fixed and repeated interactions in the context of cooperative games. However, in practical complex environments, the interac...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2003